Decomposing Centralized Multiagent Policies
نویسندگان
چکیده
Using or extending Markov decision processes (MDPs) or partially observable Markov decision processes (POMDPs) to model multiagent decision problems has become an important trend. Generally speaking, there are two types of models: centralized ones and decentralized. The centralized ones focus on finding the best joint action given any global state, while the decentralied ones try to find out that for each agent, what is the best local action given all the partial information available in that agent. Although decentralized models better capture the nature of decentralization in multiagent systems, they are much harder to solve compared to centralized models. In this paper, we show that, by studying the communication needs of the centralized models, we can establish a connection between the two models, and the solutions to centralized models (i.e. centralized policies) can be used to derive solutions to decentralized models (i.e. decentralized policies) – a process we call plan decomposition. We show that the amount of communication needed could be greatly reduced during the decomposition, and there are techniques that could be applied to produce a set of decentralized policies based on the same centralized policy. While this method does not solve decentralized models optimally, it does offer a great deal of flexibility and allows us to tradeoff the quality of the policies with the amount of communication needed, and gives us better insights about the need and timing for effective coordination in multiagent planning.
منابع مشابه
Bayesian Reinforcement Learning for Multiagent Systems with State Uncertainty
Bayesian methods for reinforcement learning are promising because they allow model uncertainty to be considered explicitly and offer a principled way of dealing with the exploration/exploitation tradeoff. However, for multiagent systems there have been few such approaches, and none of them apply to problems with state uncertainty. In this paper we fill this gap by proposing two frameworks for B...
متن کاملSolving multiagent assignment Markov decision processes
We consider the setting of multiple collaborative agents trying to complete a set of tasks as assigned by a centralized controller. We propose a scalable method called“Assignmentbased decomposition” which is based on decomposing the problem of action selection into an upper assignment level and a lower task execution level. The assignment problem is solved by search, while the task execution is...
متن کاملEfficient Offline Communication Policies for Factored Multiagent POMDPs
Factored Decentralized Partially Observable Markov Decision Processes (DecPOMDPs) form a powerful framework for multiagent planning under uncertainty, but optimal solutions require a rigid history-based policy representation. In this paper we allow inter-agent communication which turns the problem in a centralized Multiagent POMDP (MPOMDP). We map belief distributions over state factors to an a...
متن کاملMulti-Objective Optimization in Multiagent Systems
Cooperative multiagent systems are used as solution concepts in many application domains including air traffic control, satellite communications, and extra planetary exploration. As systems become more distributed and complex, we observe three phenomena. First, these systems cannot be accurately modeled, rendering traditional model based control methods inadequate. Second, system parameters are...
متن کاملMultiagent Systems Viewed as Distributed Scheduling Systems: Methodology and Experiments
In this article, we present a design technique that facilitates the work of extracting and defining the tasks scheduling problem for a multiagent system. We also compare a centralized scheduling approach to a decentralized scheduling approach to see the difference in the efficiency of the schedules and the amount of information transmitted between the agents. Our experimental results show that ...
متن کامل